📚 node [[target|target]]
Welcome! Nobody has contributed anything to 'target|target' yet. You can:
  • Write something in the document below!
    • There is at least one public document in every node in the Agora. Whatever you write in it will be integrated and made available for the next visitor to read and edit.
  • Write to the Agora from social media.
    • If you follow Agora bot on a supported platform and include the wikilink [[target|target]] in a post, the Agora will link it here and optionally integrate your writing.
  • Sign up as a full Agora user.
    • As a full user you will be able to contribute your personal notes and resources directly to this knowledge commons. Some setup required :)
⥅ related node [[2003 09 24 commsdesign voip chip targets home gateways]]
⥅ related node [[precision_target]]
⥅ related node [[target]]
⥅ related node [[target_network]]
⥅ related node [[amazon workers tell bosses to scrap constant surveillance micromanagement and unfair targets like time off task]]
⥅ node [[target]] pulled by Agora
⥅ node [[target_network]] pulled by Agora

target network

Go back to the [[AI Glossary]]

#rl

In Deep Q-learning, a neural network that is a stable approximation of the main neural network, where the main neural network implements either a Q-function or a policy. Then, you can train the main network on the Q-values predicted by the target network. Therefore, you prevent the feedback loop that occurs when the main network trains on Q-values predicted by itself. By avoiding this feedback, training stability increases.

📖 stoas
⥱ context